Prox-Method with Rate of Convergence O(1/t) for Variational Inequalities with Lipschitz Continuous Monotone Operators and Smooth Convex-Concave Saddle Point Problems
نویسنده
چکیده
Abstract. We propose a prox-type method with efficiency estimate O(2−1) for approximating saddle points of convex-concave C1,1 functions and solutions of variational inequalities with monotone Lipschitz continuous operators. Application examples include matrix games, eigenvalue minimization and computing Lovasz capacity number of a graph and are illustrated by numerical experiments with large-scale matrix games and Lovasz capacity problems.
منابع مشابه
Mirror Prox Algorithm for Multi-Term Composite Minimization and Alternating Directions
In the paper, we develop a composite version of Mirror Prox algorithm for solving convex-concave saddle point problems and monotone variational inequalities of special structure, allowing to cover saddle point/variational analogies of what is usually called “composite minimization” (minimizing a sum of an easy-to-handle nonsmooth and a general-type smooth convex functions “as if” there were no ...
متن کاملDecomposition Techniques for Bilinear Saddle Point Problems and Variational Inequalities with Affine Monotone Operators
The majority of First Order methods for large-scale convex-concave saddle point problems and variational inequalities with monotone operators are proximal algorithms which at every iteration need to minimize over problem’s domain X the sum of a linear form and a strongly convex function. To make such an algorithm practical, X should be proximal-friendly – admit a strongly convex function with e...
متن کاملSolving variational inequalities with monotone operators on domains given by Linear Minimization Oracles
The standard algorithms for solving large-scale convex–concave saddle point problems, or, more generally, variational inequalities with monotone operators, are proximal type algorithmswhich at every iteration need to compute a prox-mapping, that is, to minimize over problem’s domain X the sum of a linear form and the specific convex distance-generating function underlying the algorithms in ques...
متن کاملNon-euclidean restricted memory level method for large-scale convex optimization
We propose a new subgradient-type method for minimizing extremely large-scale nonsmooth convex functions over “simple” domains. The characteristic features of the method are (a) the possibility to adjust the scheme to the geometry of the feasible set, thus allowing to get (nearly) dimension-independent (and nearly optimal in the large-scale case) rate-of-convergence results for minimization of ...
متن کاملMirror Prox algorithm for multi-term composite minimization and semi-separable problems
In the paper, we develop a composite version of Mirror Prox algorithm for solving convexconcave saddle point problems and monotone variational inequalities of special structure, allowing to cover saddle point/variational analogies of what is usually called “composite minimization” (minimizing a sum of an easy-to-handle nonsmooth and a general-type smooth convex functions “as if” there were no n...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- SIAM Journal on Optimization
دوره 15 شماره
صفحات -
تاریخ انتشار 2004